Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
Indoor scene recognition method combined with object detection
XU Jianglang, LI Linyan, WAN Xinjun, HU Fuyuan
Journal of Computer Applications    2021, 41 (9): 2720-2725.   DOI: 10.11772/j.issn.1001-9081.2020111815
Abstract418)      PDF (1357KB)(337)       Save
In the method of combining Object detection Network (ObjectNet) and scene recognition network, the object features extracted by the ObjectNet and the scene features extracted by the scene network are inconsistent in dimensionality and property, and there is redundant information in the object features that affects the scene judgment, resulting in low recognition accuracy of scenes. To solve this problem, an improved indoor scene recognition method combined with object detection was proposed. First, the Class Conversion Matrix (CCM) was introduced into the ObjectNet to convert the object features output by ObjectNet, so that the dimension of the object features was consistent with that of the scene features, as a result, the information loss caused by inconsistency of the feature dimensions was reduced. Then, the Context Gating (CG) mechanism was used to suppress the redundant information in the features, reducing the weight of irrelevant information, and increasing the contribution of object features in scene recognition. The recognition accuracy of the proposed method on MIT Indoor67 dataset reaches 90.28%, which is 0.77 percentage points higher than that of Spatial-layout-maintained Object Semantics Features (SOSF) method; and the recognition accuracy of the proposed method on SUN397 dataset is 81.15%, which is 1.49 percentage points higher than that of Hierarchy of Alternating Specialists (HoAS) method. Experimental results show that the proposed method improves the accuracy of indoor scene recognition.
Reference | Related Articles | Metrics
Text-to-image synthesis method based on multi-level structure generative adversarial networks
SUN Yu, LI Linyan, YE Zihan, HU Fuyuan, XI Xuefeng
Journal of Computer Applications    2019, 39 (11): 3204-3209.   DOI: 10.11772/j.issn.1001-9081.2019051077
Abstract455)      PDF (1012KB)(529)       Save
In recent years, the Generative Adversarial Network (GAN) has achieved remarkable success in text-to-image synthesis, but there are still problems such as edge blurring of images, unclear local textures, small sample variance. In view of the above shortcomings, based on Stack Generative Adversarial Network model (StackGAN++), a Multi-Level structure Generative Adversarial Networks (MLGAN) model was proposed, which is composed of multiple generators and discriminators in a hierarchical structure. Firstly, hierarchical structure coding method and word vector constraint were introduced to change the condition vector of generator of each level in the network, so that the edge details and local textures of the image were clearer and more vivid. Then, the generator and the discriminator were jointed by trained to approximate the real image distribution by using the generated image distribution of multiple levels, so that the variance of the generated sample became larger, and the diversity of the generated sample was increased. Finally, different scale images of the corresponding text were generated by generators of different levels. The experimental results show that the Inception scores of the MLGAN model reached 4.22 and 3.88 respectively on CUB and Oxford-102 datasets, which were respectively 4.45% and 3.74% higher than that of StackGAN++. The MLGAN model has improvement in solving edge blurring and unclear local textures of the generated image, and the image generated by the model is closer to the real image.
Reference | Related Articles | Metrics
Unsupervised cross-domain transfer network for 3D/2D registration in surgical navigation
WANG Xiyuan, ZHANG Zhancheng, XU Shaokang, ZHANG Baocheng, LUO Xiaoqing , HU Fuyuan
Journal of Computer Applications    DOI: 10.11772/j.issn.1001-9081.2023091332
Online available: 31 January 2024